Spectral gradient methods for linearly constrained optimization
نویسندگان
چکیده
Linearly constrained optimization problems with simple bounds are considered in the present work. First, a preconditioned spectral gradient method is defined for the case in which no simple bounds are present. This algorithm can be viewed as a quasiNewton method in which the approximate Hessians satisfy a weak secant equation. The spectral choice of steplength is embedded into the Hessian approximation, and the whole process is combined with a nonmonotone line search strategy. The simple bounds are then taken into account by placing them in an exponential penalty term that modifies the objective function. The exponential penalty scheme defines the outer iterations of the process. Each outer iteration involves the application of the previously defined preconditioned spectral gradient method for linear equality constrained problems. Therefore, an equality constrained convex quadratic programming problem needs to be solved at every inner iteration. The associated extended KKT matrix remains constant unless the process is reinitiated. In ordinary inner iterations, only the right hand side of the KKT system changes. Therefore, suitable sparse factorization techniques can be effectively applied and exploited. Encouraging numerical experiments are presented.
منابع مشابه
Spectral projected gradient and variable metric methods for optimization with linear inequalities
A family of variable metric methods for convex constrained optimization was introduced recently by Birgin, Mart́ınez and Raydan. One of the members of this family is the Inexact Spectral Projected Gradient (ISPG) method for minimization with convex constraints. At each iteration of these methods a strictly convex quadratic function with convex constraints must be (inexactly) minimized. In the ca...
متن کاملPattern Search Methods for Linearly Constrained Minimization
We extend pattern search methods to linearly constrained minimization. We develop a general class of feasible point pattern search algorithms and prove global convergence to a KarushKuhn-Tucker point. As in the case of unconstrained minimization, pattern search methods for linearly constrained problems accomplish this without explicit recourse to the gradient or the directional derivative of th...
متن کاملOn the hybrid conjugate gradient method for solving fuzzy optimization problem
In this paper we consider a constrained optimization problem where the objectives are fuzzy functions (fuzzy-valued functions). Fuzzy constrained Optimization (FO) problem plays an important role in many fields, including mathematics, engineering, statistics and so on. In the other side, in the real situations, it is important to know how may obtain its numerical solution of a given interesting...
متن کاملA Derivative-free Method for Linearly Constrained Nonsmooth Optimization
This paper develops a new derivative-free method for solving linearly constrained nonsmooth optimization problems. The objective functions in these problems are, in general, non-regular locally Lipschitz continuous function. The computation of generalized subgradients of such functions is difficult task. In this paper we suggest an algorithm for the computation of subgradients of a broad class ...
متن کاملEnhancing Sparsity by Constraining Strategies: Constrained SIRT versus Spectral Projected Gradient Methods
We investigate a constrained version of simultaneous iterative reconstruction techniques (SIRT) from the general viewpoint of projected gradient methods. This connection enable us to assess the computational merit of this algorithm class. We borrow a leaf from numerical optimization to cope with the slow convergence of projected gradient methods and propose an acceleration procedure based on th...
متن کامل